IV Regularization Tools for TrainingLarge - Scale Neural Networks
نویسنده
چکیده
We present regularization tools for training small-and-medium as well as large-scale artiicial feedforward neural networks. The determination of the weights leads to very ill-conditioned nonlinear least squares problems and regularization is often suggested to get control over the network complexity, small variance error, and nice optimization problems. The algorithms proposed solve explicitly a sequence of Tikhonov regularized nonlinear least squares problems. For small-and-medium size problems the Gauss-Newton method is applied to the regularized problem that is much more well-conditioned than the original problem, and exhibits far better convergence properties than a Levenberg-Marquardt method. Numerical results presented also connrm that the proposed implementations are more reliable and eecient than the Levenberg-Marquardt method. For large-scale problems, methods using new special purpose automatic diierentiation combined with conjugate gradient methods are proposed. The algorithms developed utilize the structure of the problem in diierent ways and perform much better than Polak-Ribi ere based methods. All algorithms are tested using benchmark problems and guidelines by Lutz Prechelt in the Proben1 package. All software is programmed in Matlab and gathered in a toolbox.
منابع مشابه
Training Large Neural Networks
We describe regularization tools for training large-scale artiicial feed-forward neural networks. We propose algorithms that explicitly use a sequence of Tikhonov regularized nonlinear least squares problems. For large-scale problems, methods using new special purpose automatic diierentiation are used in a conjugate gradient method for computing a truncated Gauss-Newton search direction. The al...
متن کاملRegularization for Unsupervised Deep Neural Nets
Unsupervised neural networks, such as restricted Boltzmann machines (RBMs) and deep belief networks (DBNs), are powerful tools for feature selection and pattern recognition tasks. We demonstrate that overfitting occurs in such models just as in deep feedforward neural networks, and discuss possible regularization methods to reduce overfitting. We also propose a “partial” approach to improve the...
متن کاملA Survey of Forecasting Preprocessing Techniques using RNs
In this paper we make a survey of various preprocessing techniques including the statistical method for volatile time series forecasting using Regularization Networks (RNs). These methods improve the performance of Regularization Networks i.e. using Independent Component Analysis (ICA) algorithms and filtering as preprocessing tools. The preprocessed data is introduced into a Regularized Artifi...
متن کاملPublishersREGULARIZATION TOOLS FOR TRAININGFEED - FORWARD NEURAL NETWORKSPART II : Large
We describe regularization tools for training large-scale artiicial feed-forward neural networks. In a companion paper (in this issue) we give the basic ideas and some theoretical results regarding the Gauss-Newton method compared to other methods such as the Levenberg-Marquardt method applied on small and medium size problems. We propose algorithms that explicitly use a sequence of Tikhonov re...
متن کاملRegularization of Neural
Neural networks are exible tools for nonlinear function approximation and by expanding the network any relevant target function can be approximated 6]. The risk of overrtting on noisy data is of major concern in neural network design 2]. By using regularization, overrtting is reduced, thereby improving generalization ability on future data. In this contribution we present a scheme for estimatio...
متن کامل